Learning from Imprecise and Fuzzy Observations: Data Disambiguation through Generalized Loss Minimization

نویسنده

  • Eyke Hüllermeier
چکیده

Methods for analyzing or learning from “fuzzy data” have attracted increasing attention in recent years. In many cases, however, existing methods (for precise, non-fuzzy data) are extended to the fuzzy case in an ad-hoc manner, and without carefully considering the interpretation of a fuzzy set when being used for modeling data. Distinguishing between an ontic and an epistemic interpretation of fuzzy set-valued data, and focusing on the latter, we argue that a “fuzzification” of learning algorithms based on an application of the generic extension principle is not appropriate. In fact, the extension principle fails to properly exploit the inductive bias underlying statistical and machine learning methods, although this bias, at least in principle, offers a means for “disambiguating” the fuzzy data. Alternatively, we therefore propose a method which is based on the generalization of loss functions in empirical risk minimization, and which performs model identification and data disambiguation simultaneously. Elaborating on the fuzzification of specific types of losses, we establish connections to well-known loss functions in regression and classification. We compare our approach with related methods and illustrate its use in logistic regression for binary classification.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Superset Learning Based on Generalized Loss Minimization

In standard supervised learning, each training instance is associated with an outcome from a corresponding output space (e.g., a class label in classification or a real number in regression). In the superset learning problem, the outcome is only characterized in terms of a superset—a subset of candidates that covers the true outcome but may also contain additional ones. Thus, superset learning ...

متن کامل

Fuzzy clustering with the generalized entropy of feature weights

Fuzzy c-means (FCM) is an important clustering algorithm. However, it does not consider the impact of different feature on clustering. In this paper, we present a fuzzy clustering algorithm with the generalized entropy of feature weights FCM (GEWFCM). By introducing feature weights and adding regularized term of their generalized entropy, a new objective function is proposed in terms of objecti...

متن کامل

Generalized Fuzzy Inverse Data envelopment Analysis Models

Traditional DEA models do not deal with imprecise data and assume that the data for all inputs and outputs are known exactly. Inverse DEA models can be used to estimate inputs for a DMU when some or all outputs and efficiency level of this DMU are increased or preserved. this paper studies the inverse DEA for fuzzy data. This paper proposes generalized inverse DEA in fuzzy data envelopment anal...

متن کامل

Linguistic cost-sensitive learning of genetic fuzzy classifiers for imprecise data

Cost-sensitive classification is based on a set of weights defining the expected cost of misclassifying an object. In this paper, a Genetic Fuzzy Classifier, which is able to extract fuzzy rules from interval or fuzzy valued data, is extended to this type of classification. This extension consists in enclosing the estimation of the expected misclassification risk of a classifier, when assessed ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Int. J. Approx. Reasoning

دوره 55  شماره 

صفحات  -

تاریخ انتشار 2014